On the Computational Efficiency of Training Neural Networks
نویسندگان
چکیده
It is well-known that neural networks are computationally hard to train. On the other hand, in practice, modern day neural networks are trained efficiently using SGD and a variety of tricks that include different activation functions (e.g. ReLU), over-specification (i.e., train networks which are larger than needed), and regularization. In this paper we revisit the computational complexity of training neural networks from a modern perspective. We provide both positive and negative results, some of them yield new provably efficient and practical algorithms for training certain types of neural networks.
منابع مشابه
The relationship between Neural Networks and DEA-R (Case Study: Companies Stock Exchange)
Evaluate the performance of companies on the Stock Exchange using non-parametric methods is very important. DEA and DEA-R with the strategies for piecewise linear frontier production function and use of available data, assess the stock company. In this study, using a neural network algorithm DEA and DEA-R is suggested to classify the first companies in the stock exchange; Secondly, using the...
متن کاملA conjugate gradient based method for Decision Neural Network training
Decision Neural Network is a new approach for solving multi-objective decision-making problems based on artificial neural networks. Using inaccurate evaluation data, network training has improved and the number of educational data sets has decreased. The available training method is based on the gradient decent method (BP). One of its limitations is related to its convergence speed. Therefore,...
متن کاملOn the use of back propagation and radial basis function neural networks in surface roughness prediction
Various artificial neural networks types are examined and compared for the prediction of surface roughness in manufacturing technology. The aim of the study is to evaluate different kinds of neural networks and observe their performance and applicability on the same problem. More specifically, feed-forward artificial neural networks are trained with three different back propagation algorithms, ...
متن کاملPREDICTION OF NONLINEAR TIME HISTORY DEFLECTION OF SCALLOP DOMES BY NEURAL NETWORKS
This study deals with predicting nonlinear time history deflection of scallop domes subject to earthquake loading employing neural network technique. Scallop domes have alternate ridged and grooves that radiate from the centre. There are two main types of scallop domes, lattice and continuous, which the latticed type of scallop domes is considered in the present paper. Due to the large number o...
متن کاملآموزش شبکه عصبی MLP در فشردهسازی تصاویر با استفاده از روش GSA
Image compression is one of the important research fields in image processing. Up to now, different methods are presented for image compression. Neural network is one of these methods that has represented its good performance in many applications. The usual method in training of neural networks is error back propagation method that its drawbacks are late convergence and stopping in points of lo...
متن کاملWavelet Neural Network with Random Wavelet Function Parameters
The training algorithm of Wavelet Neural Networks (WNN) is a bottleneck which impacts on the accuracy of the final WNN model. Several methods have been proposed for training the WNNs. From the perspective of our research, most of these algorithms are iterative and need to adjust all the parameters of WNN. This paper proposes a one-step learning method which changes the weights between hidden la...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014